1,604 research outputs found

    New Physics Agnostic Selections For New Physics Searches

    Get PDF
    We discuss a model-independent strategy for boosting new physics searches with the help of an unsupervised anomaly detection algorithm. Prior to a search, each input event is preprocessed by the algorithm - a variational autoencoder (VAE). Based on the loss assigned to each event, input data can be split into a background control sample and a signal enriched sample. Following this strategy, one can enhance the sensitivity to new physics with no assumption on the underlying new physics signature. Our results show that a typical BSM search on the signal enriched group is more sensitive than an equivalent search on the original dataset

    FPGA-accelerated machine learning inference as a service for particle physics computing

    Full text link
    New heterogeneous computing paradigms on dedicated hardware with increased parallelization, such as Field Programmable Gate Arrays (FPGAs), offer exciting solutions with large potential gains. The growing applications of machine learning algorithms in particle physics for simulation, reconstruction, and analysis are naturally deployed on such platforms. We demonstrate that the acceleration of machine learning inference as a web service represents a heterogeneous computing solution for particle physics experiments that potentially requires minimal modification to the current computing model. As examples, we retrain the ResNet-50 convolutional neural network to demonstrate state-of-the-art performance for top quark jet tagging at the LHC and apply a ResNet-50 model with transfer learning for neutrino event classification. Using Project Brainwave by Microsoft to accelerate the ResNet-50 image classification model, we achieve average inference times of 60 (10) milliseconds with our experimental physics software framework using Brainwave as a cloud (edge or on-premises) service, representing an improvement by a factor of approximately 30 (175) in model inference latency over traditional CPU inference in current experimental hardware. A single FPGA service accessed by many CPUs achieves a throughput of 600--700 inferences per second using an image batch of one, comparable to large batch-size GPU throughput and significantly better than small batch-size GPU throughput. Deployed as an edge or cloud service for the particle physics computing model, coprocessor accelerators can have a higher duty cycle and are potentially much more cost-effective.Comment: 16 pages, 14 figures, 2 table

    Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors

    Full text link
    Although the quest for more accurate solutions is pushing deep learning research towards larger and more complex algorithms, edge devices demand efficient inference and therefore reduction in model size, latency and energy consumption. One technique to limit model size is quantization, which implies using fewer bits to represent weights and biases. Such an approach usually results in a decline in performance. Here, we introduce a method for designing optimally heterogeneously quantized versions of deep neural network models for minimum-energy, high-accuracy, nanosecond inference and fully automated deployment on chip. With a per-layer, per-parameter type automatic quantization procedure, sampling from a wide range of quantizers, model energy consumption and size are minimized while high accuracy is maintained. This is crucial for the event selection procedure in proton-proton collisions at the CERN Large Hadron Collider, where resources are strictly limited and a latency of O(1) μ{\mathcal O}(1)~\mus is required. Nanosecond inference and a resource consumption reduced by a factor of 50 when implemented on field-programmable gate array hardware are achieved

    Fast convolutional neural networks on FPGAs with hls4ml

    Get PDF
    We introduce an automated tool for deploying ultra low-latency, low-power deep neural networks with convolutional layers on FPGAs. By extending the hls4ml library, we demonstrate an inference latency of 5μ5\,\mus using convolutional architectures, targeting microsecond latency applications like those at the CERN Large Hadron Collider. Considering benchmark models trained on the Street View House Numbers Dataset, we demonstrate various methods for model compression in order to fit the computational constraints of a typical FPGA device used in trigger and data acquisition systems of particle detectors. In particular, we discuss pruning and quantization-aware training, and demonstrate how resource utilization can be significantly reduced with little to no loss in model accuracy. We show that the FPGA critical resource consumption can be reduced by 97% with zero loss in model accuracy, and by 99% when tolerating a 6% accuracy degradation.Comment: 18 pages, 18 figures, 4 table

    Search for supersymmetry in pp collisions at s\sqrt{s} = 13 tev in the single-lepton final state using the sum of masses of large-radius jets

    Full text link
    Results are reported from a search for supersymmetric particles in proton-proton collisions in the final state with a single lepton, multiple jets, including at least one b-tagged jet, and large missing transverse momentum. The search uses a sample of proton-proton collision data at s\sqrt{s} = 13 TeV recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 35.9  fb1fb^{−1}. The observed event yields in the signal regions are consistent with those expected from standard model backgrounds. The results are interpreted in the context of simplified models of supersymmetry involving gluino pair production, with gluino decay into either on- or off-mass-shell top squarks. Assuming that the top squarks decay into a top quark plus a stable, weakly interacting neutralino, scenarios with gluino masses up to about 1.9 TeV are excluded at 95% confidence level for neutralino masses up to about 1 TeV

    Search for dark matter produced in association with heavy-flavor quark pairs in proton-proton collisions at s\sqrt{s} = 13 TeV

    Full text link
    A search is presented for an excess of events with heavy-flavor quark pairs (ttt\overline{t} and bbb\overline{b}) and a large imbalance in transverse momentum in data from proton–proton collisions at a center-of-mass energy of 13TeV. The data correspond to an integrated luminosity of 2.2 fb1fb^{−1} collected with the CMS detector at the CERN LHC. No deviations are observed with respect to standard model predictions. The results are used in the first interpretation of dark matter production in ttt\overline{t} and bbb\overline{b} final states in a simplified model. This analysis is also the first to perform a statistical combination of searches for dark matter produced with different heavy-flavor final states. The combination provides exclusions that are stronger than those achieved with individual heavy-flavor final states

    Combination of searches for heavy resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collisions at s\sqrt{s} = 8 and 13 TeV

    Full text link
    A statistical combination of searches is presented for massive resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collision data collected by the CMS experiment at the LHC. The data were taken at centre-of-mass energies of 8 and 13 TeV, corresponding to respective integrated luminosities of 19.7 and up to 2.7 fb1fb^{−1}. The results are interpreted in the context of heavy vector triplet and singlet models that mimic properties of composite-Higgs models predicting W′ and Z′ bosons decaying to WZ, WW, WH, and ZH bosons. A model with a bulk graviton that decays into WW and ZZ is also considered. This is the first combined search for WW, WZ, WH, and ZH resonances and yields lower limits on masses at 95% confidence level for W′ and Z′ singlets at 2.3 TeV, and for a triplet at 2.4 TeV. The limits on the production cross section of a narrow bulk graviton resonance with the curvature scale of the warped extra dimension k~=0.5\tilde{k}=0.5, in the mass range of 0.6 to 4.0 TeV, are the most stringent published to date

    The CMS trigger system

    Full text link
    This paper describes the CMS trigger system and its performance during Run 1 of the LHC. The trigger system consists of two levels designed to select events of potential physics interest from a GHz (MHz) interaction rate of proton-proton (heavy ion) collisions. The first level of the trigger is implemented in hardware, and selects events containing detector signals consistent with an electron, photon, muon, τ lepton, jet, or missing transverse energy. A programmable menu of up to 128 object-based algorithms is used to select events for subsequent processing. The trigger thresholds are adjusted to the LHC instantaneous luminosity during data taking in order to restrict the output rate to 100 kHz, the upper limit imposed by the CMS readout electronics. The second level, implemented in software, further refines the purity of the output stream, selecting an average rate of 400 Hz for offline event storage. The objectives, strategy and performance of the trigger system during the LHC Run 1 are described
    corecore